Goto

Collaborating Authors

 agentic ai


What is Moltbook? The strange new social media site for AI bots

The Guardian

Some people are sceptical about whether the socialising of bots is a sign of what is coming with the rise of agentic AI. Some people are sceptical about whether the socialising of bots is a sign of what is coming with the rise of agentic AI. A bit like Reddit for artificial intelligence, Moltbook allows AI agents - bots built by humans - to post and interact with each other. On social media, people often accuse each other of being bots, but what happens when an entire social network is designed for AI agents to use? Moltbook is a site where the AI agents - bots built by humans - can post and interact with each other. It is designed to look like Reddit, with subreddits on different topics and upvoting.


What's coming up at #AAAI2026?

AIHub

We (AIhub) will be running a short course on science communication on Wednesday 21 January, from 13:00 - 14:30. In this brief tutorial, science communication experts will teach you how to clearly and concisely explain your research to non-specialists.


David vs. Goliath: Can Small Models Win Big with Agentic AI in Hardware Design?

Shankar, Shashwat, Pandey, Subhranshu, Mochahari, Innocent Dengkhw, Mali, Bhabesh, Chowdhury, Animesh Basak, Bhattacharjee, Sukanta, Karfa, Chandan

arXiv.org Artificial Intelligence

Large Language Model(LLM) inference demands massive compute and energy, making domain-specific tasks expensive and unsustainable. As foundation models keep scaling, we ask: Is bigger always better for hardware design? Our work tests this by evaluating Small Language Models coupled with a curated agentic AI framework on NVIDIA's Comprehensive Verilog Design Problems(CVDP) benchmark. Results show that agentic workflows: through task decomposition, iterative feedback, and correction - not only unlock near-LLM performance at a fraction of the cost but also create learning opportunities for agents, paving the way for efficient, adaptive solutions in complex design tasks.


STRIDE: A Systematic Framework for Selecting AI Modalities -- Agentic AI, AI Assistants, or LLM Calls

Asthana, Shubhi, Zhang, Bing, DeLuca, Chad, Mahindru, Ruchi, Patel, Hima

arXiv.org Artificial Intelligence

The rapid shift from stateless large language models (LLMs) to autonomous, goal-driven agents raises a central question: When is agentic AI truly necessary? While agents enable multi-step reasoning, persistent memory, and tool orchestration, deploying them indiscriminately leads to higher cost, complexity, and risk. We present STRIDE (Systematic Task Reasoning Intelligence Deployment Evaluator), a framework that provides principled recommendations for selecting between three modalities: (i) direct LLM calls, (ii) guided AI assistants, and (iii) fully autonomous agentic AI. STRIDE integrates structured task decomposition, dynamism attribution, and self-reflection requirement analysis to produce an Agentic Suitability Score, ensuring that full agentic autonomy is reserved for tasks with inherent dynamism or evolving context. Evaluated across 30 real-world tasks spanning SRE, compliance, and enterprise automation, STRIDE achieved 92% accuracy in modality selection, reduced unnecessary agent deployments by 45%, and cut resource costs by 37%. Expert validation over six months in SRE and compliance domains confirmed its practical utility, with domain specialists agreeing that STRIDE effectively distinguishes between tasks requiring simple LLM calls, guided assistants, or full agentic autonomy. This work reframes agent adoption as a necessity-driven design decision, ensuring autonomy is applied only when its benefits justify the costs.


Semantic Trading: Agentic AI for Clustering and Relationship Discovery in Prediction Markets

Capponi, Agostino, Gliozzo, Alfio, Zhu, Brian

arXiv.org Artificial Intelligence

Prediction markets allow users to trade on outcomes of real-world events, but are prone to fragmentation with overlapping questions, implicit equivalences, and hidden contradictions across markets. We present an agentic AI pipeline that autonomously (i) clusters markets into coherent topical groups using natural-language understanding over contract text and metadata, and (ii) identifies within-cluster market pairs whose resolved outcomes exhibit strong dependence, including "same-outcome" (correlated) and "different-outcome" (anti-correlated) relationships. Using a historical dataset of resolved markets on Poly-market, we evaluate the accuracy of the agent's relational predictions. We then synthesize discovered relationships into a simple trading strategy to quantify how discovered relationships translate into actionable strategies. Results show that agent-identified relationships have around 60-70% accuracy, and their induced trading strategies have an average return of 20% over week-long horizons, highlighting the ability of agen-tic AI and large language models to uncover latent semantic structure within prediction markets.


Agentifying Agentic AI

Dignum, Virginia, Dignum, Frank

arXiv.org Artificial Intelligence

Agentic AI seeks to endow systems with sustained autonomy, reasoning, and interaction capabilities. To realize this vision, its assumptions about agency must be complemented by explicit models of cognition, cooperation, and governance. This paper argues that the conceptual tools developed within the Autonomous Agents and Multi-Agent Systems (AAMAS) community, such as BDI architectures, communication protocols, mechanism design, and institutional modelling, provide precisely such a foundation. By aligning adaptive, data-driven approaches with structured models of reasoning and coordination, we outline a path toward agentic systems that are not only capable and flexible, but also transparent, cooperative, and accountable. The result is a perspective on agency that bridges formal theory and practical autonomy.


Designing digital resilience in the agentic AI era

MIT Technology Review

As AI shifts from leveraging information provided by humans to making decisions on their behalf, tech leaders must weave an intelligent data fabric to unlock the full potential of agentic AI while shoring up enterprise-wide resilience. Digital resilience--the ability to prevent, withstand, and recover from digital disruptions--has long been a strategic priority for enterprises. With the rise of agentic AI, the urgency for robust resilience is greater than ever. Agentic AI represents a new generation of autonomous systems capable of proactive planning, reasoning, and executing tasks with minimal human intervention. As these systems shift from experimental pilots to core elements of business operations, they offer new opportunities but also introduce new challenges when it comes to ensuring digital resilience. That's because the autonomy, speed, and scale at which agentic AI operates can amplify the impact of even minor data inconsistencies, fragmentation, or security gaps.


Enhancing Demand-Oriented Regionalization with Agentic AI and Local Heterogeneous Data for Adaptation Planning

Noorani, Seyedeh Mobina, Gao, Shangde, Chen, Changjie, Ochoa, Karla Saldana

arXiv.org Artificial Intelligence

Conventional planning units or urban regions, such as census tracts, zip codes, or neighborhoods, often do not capture the specific demands of local communities and lack the flexibility to implement effective strategies for hazard prevention or response. To support the creation of dynamic planning units, we introduce a planning support system with agentic AI that enables users to generate demand-oriented regions for disaster planning, integrating the human-in-the-loop principle for transparency and adaptability. The platform is built on a representative initialized spatially constrained self-organizing map (RepSC-SOM), extending traditional SOM with adaptive geographic filtering and region-growing refinement, while AI agents can reason, plan, and act to guide the process by suggesting input features, guiding spatial constraints, and supporting interactive exploration. We demonstrate the capabilities of the platform through a case study on the flooding-related risk in Jacksonville, Florida, showing how it allows users to explore, generate, and evaluate regionalization interactively, combining computational rigor with user-driven decision making.


Improving VMware migration workflows with agentic AI

MIT Technology Review

As licensing costs surge and cloud use becomes more strategic, AI agents are turning months of manual migration work for IT teams into weeks of machine-assisted automation. For years, many chief information officers (CIOs) looked at VMware-to-cloud migrations with a wary pragmatism. Manually mapping dependencies and rewriting legacy apps mid-flight was not an enticing, low-lift proposition for enterprise IT teams. But the calculus for such decisions has changed dramatically in a short period of time. Following recent VMware licensing changes, organizations are seeing greater uncertainty around the platform's future. At the same time, cloud-native innovation is accelerating.


From Failure Modes to Reliability Awareness in Generative and Agentic AI System

Janet, null, Lin, null, Zhang, Liangwei

arXiv.org Artificial Intelligence

This chapter bridges technical analysis and organizational preparedness by tracing the path from layered failure modes to reliability awareness in generative and agentic AI systems. We first introduce an 11-layer failure stack, a structured framework for identifying vulnerabilities ranging from hardware and power foundations to adaptive learning and agentic reasoning. Building on this, the chapter demonstrates how failures rarely occur in isolation but propagate across layers, creating cascading effects with systemic consequences. To complement this diagnostic lens, we develop the concept of awareness mapping: a maturity-oriented framework that quantifies how well individuals and organizations recognize reliability risks across the AI stack. Awareness is treated not only as a diagnostic score but also as a strategic input for AI governance, guiding improvement and resilience planning. By linking layered failures to awareness levels and further integrating this into Dependability-Centred Asset Management (DCAM), the chapter positions awareness mapping as both a measurement tool and a roadmap for trustworthy and sustainable AI deployment across mission-critical domains.